Local Minimax Learning of Approximately Polynomial Functions

نویسندگان

  • Lee Jones
  • Konstantin A. Rybnikov
چکیده

Suppose we have a number of noisy measurements of an unknown real-valued function f near a point of interest x0 ∈ R. Suppose also that nothing can be assumed about the noise distribution, except for zero mean and bounded covariance matrix. We want to estimate f at x0 using a general linear parametric family f(x; a) = a0h0(x) + . . . + aqhq(x), where a ∈ R and hi’s are bounded functions on a neighborhood B of x0, which contains all points of measurement. Typically, B is a Euclidean ball or cube in R (more generally, a ball in an lp-norm). In the case when the hi’s are polynomial functions in (x1, . . . , xd) = x the model is called locally-polynomial. In particular, if the hi’s form a basis of the linear space of polynomials of degree at most two, the model is called locally-quadratic (if the degree is at most three, the model is locally-cubic, etc.). More generally, hi are picked from some linear space of functions H, which must include at least all affine functions on R. Often, information about the behavior of function f on B is available, which is referred to as context, such as, e.g., that f takes values in a known interval, or that it satisfies a Lipschitz condition, etc. Given a loss function and a linear space of functions H, the idea of local minimax learning is in choosing for each point of interest x0 a parameter vector a that minimizes the maximal possible loss over all a ∈ R. The bounds and algorithms are not based on asymptotics or Bayesian assumptions and are truly local for each query, not depending on cross validating estimates at other queries to optimize modeling. The theory of local minimax estimation with context for locally-polynomial models and approximately locally polynomial models has been recently initiated by Jones (2006) and the focus of our paper is on a subclass of problems studied by Jones, which reduce to real algebraic geometry and finite-dimensional optimization of linear functions over compact domains. See Jones (2006) for detailed treatment. Denote by Pr,d the space of real polynomial of degree at most r in d indeterminates. In the case of H = P1,d and a given bound on the change of f on B = {x ∈ R |x| ≤ 1}, the solution for the squared error loss function is in the form of ridge regression, where the ridge parameter is identified; hence, a minimax justification for ridge regression is given, together with explicit best error bounds. The analysis of polynomial models of degree above 1 leads to interesting and difficult questions in real algebraic geometry and non-linear optimization. We show that in the

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

On Automatic Boundary Corrections*

Many popular curve estimators based on smoothing have diicul-ties caused by boundary eeects. These eeects are visually disturbing in practice and can play a dominant role in theoretical analysis. Local polynomial regression smoothers are known to correct boundary eeects automatically. Some analogs are implemented for density estimation and the resulting estimators also achieve automatic boundar...

متن کامل

On the Minimax Optimality of Block Thresholded Wavelets Estimators for ?-Mixing Process

We propose a wavelet based regression function estimator for the estimation of the regression function for a sequence of ?-missing random variables with a common one-dimensional probability density function. Some asymptotic properties of the proposed estimator based on block thresholding are investigated. It is found that the estimators achieve optimal minimax convergence rates over large class...

متن کامل

Minimax Estimation of the Scale Parameter in a Family of Transformed Chi-Square Distributions under Asymmetric Squared Log Error and MLINEX Loss Functions

This paper is concerned with the problem of finding the minimax estimators of the scale parameter ? in a family of transformed chi-square distributions, under asymmetric squared log error (SLE) and modified linear exponential (MLINEX) loss functions, using the Lehmann Theorem [2]. Also we show that the results of Podder et al. [4] for Pareto distribution are a special case of our results for th...

متن کامل

Optimal Rates for the Regularized Learning Algorithms under General Source Condition

We consider the learning algorithms under general source condition with the polynomial decay of the eigenvalues of the integral operator in vector-valued function setting. We discuss the upper convergence rates of Tikhonov regularizer under general source condition corresponding to increasing monotone index function. The convergence issues are studied for general regularization schemes by using...

متن کامل

ar X iv : 1 61 1 . 01 90 0 v 1 [ st at . M L ] 7 N ov 2 01 6 Optimal rates for the regularized learning algorithms under general source condition

We consider the learning algorithms under general source condition with the polynomial decay of the eigenvalues of the integral operator in vector-valued function setting. We discuss the upper convergence rates of Tikhonov regularizer under general source condition corresponding to increasing monotone index function. The convergence issues are studied for general regularization schemes by using...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006